Tompkins County
G-OSR: A Comprehensive Benchmark for Graph Open-Set Recognition
Dong, Yicong, He, Rundong, Chen, Guangyao, Zhang, Wentao, Han, Zhongyi, Shi, Jieming, Yin, Yilong
--Graph Neural Networks (GNNs) have achieved significant success in machine learning, with wide applications in social networks, bioinformatics, knowledge graphs, and other fields. Most research assumes ideal closed-set environments. However, in real-world open-set environments, graph learning models face challenges in robustness and reliability due to unseen classes. This highlights the need for Graph Open-Set Recognition (GOSR) methods to address these issues and ensure effective GNN application in practical scenarios. Research in GOSR is in its early stages, with a lack of a comprehensive benchmark spanning diverse tasks and datasets to evaluate methods. Moreover, traditional methods, Graph Out-of-Distribution Detection (GOODD), GOSR, and Graph Anomaly Detection (GAD) have mostly evolved in isolation, with little exploration of their interconnections or potential applications to GOSR. T o fill these gaps, we introduce G-OSR, a comprehensive benchmark for evaluating GOSR methods at both the node and graph levels, using datasets from multiple domains to ensure fair and standardized comparisons of effectiveness and efficiency across traditional, GOODD, GOSR, and GAD methods. The results offer critical insights into the generalizability and limitations of current GOSR methods and provide valuable resources for advancing research in this field through systematic analysis of diverse approaches. RAPH learning, as a significant research direction in machine learning, has been widely applied in social network analysis, recommendation systems, bioinformatics, knowledge graphs, traffic planning, and the fields of chemistry and materials science [1]. Graph Neural Networks (GNNs) have demonstrated superior performance in various node classification and graph classification tasks [2]. These methods typically follow a closed-set setting, which assumes that all test classes are among the seen classes accessible during training [3]. However, in real-world scenarios, due to undersampling, out-of-distribution, or anomalous samples, it is highly likely to encounter samples belonging to novel unseen classes, which can significantly impact the safety and robustness of models [4], as illustrated in Figure 1. Guangyao Chen is with Cornell University, Ithaca, NY, USA. Wentao Zhang is with Peking University, Beijing, China. Zhongyi Han is with King Abdullah University of Science and Technology, Thuwal, Saudi Arabia. Rundong He and Yilong Yin are the corresponding authors. Closed-set classification cannot identify unseen classes, while open-set recognition can identify unseen classes and classify nodes belonging to seen classes.
CoCoNUT: Structural Code Understanding does not fall out of a tree
Large Language Models (LLMs) have shown impressive performance across a wide array of tasks involving both structured and unstructured textual data. Recent results on various benchmarks for code generation, repair, or completion suggest that certain models have programming abilities comparable to or even surpass humans. In this work, we demonstrate that high performance on such benchmarks does not correlate to humans' innate ability to understand structural control flow in code. To this end, we extract solutions from the HumanEval benchmark, which the relevant models perform strongly on, and trace their execution path using function calls sampled from the respective test set. Using this dataset, we investigate the ability of seven state-of-the-art LLMs to match the execution trace and find that, despite their ability to generate semantically identical code, they possess limited ability to trace execution paths, especially for longer traces and specific control structures. We find that even the top-performing model, Gemini, can fully and correctly generate only 47% of HumanEval task traces. Additionally, we introduce a subset for three key structures not contained in HumanEval: Recursion, Parallel Processing, and Object-Oriented Programming, including concepts like Inheritance and Polymorphism. Besides OOP, we show that none of the investigated models achieve an accuracy over 5% on the relevant traces. Aggregating these specialized parts with HumanEval tasks, we present CoCoNUT: Code Control Flow for Navigation Understanding and Testing, which measures a model's ability to trace execution of code upon relevant calls, including advanced structural components. We conclude that current LLMs need significant improvement to enhance code reasoning abilities. We hope our dataset helps researchers bridge this gap.
A Regularized Framework for Sparse and Structured Neural Attention Mathieu Blondel Cornell University NTT Communication Science Laboratories Ithaca, NY
Modern neural networks are often augmented with an attention mechanism, which tells the network where to focus within the input. We propose in this paper a new framework for sparse and structured attention, building upon a smoothed max operator. We show that the gradient of this operator defines a mapping from real values to probabilities, suitable as an attention mechanism. Our framework includes softmax and a slight generalization of the recently-proposed sparsemax as special cases. However, we also show how our framework can incorporate modern structured penalties, resulting in more interpretable attention mechanisms, that focus on entire segments or groups of an input. We derive efficient algorithms to compute the forward and backward passes of our attention mechanisms, enabling their use in a neural network trained with backpropagation. To showcase their potential as a drop-in replacement for existing ones, we evaluate our attention mechanisms on three large-scale tasks: textual entailment, machine translation, and sentence summarization. Our attention mechanisms improve interpretability without sacrificing performance; notably, on textual entailment and summarization, we outperform the standard attention mechanisms based on softmax and sparsemax.
BirdSet: A Dataset and Benchmark for Classification in Avian Bioacoustics
Rauch, Lukas, Schwinger, Raphael, Wirth, Moritz, Heinrich, René, Huseljic, Denis, Lange, Jonas, Kahl, Stefan, Sick, Bernhard, Tomforde, Sven, Scholz, Christoph
Deep learning (DL) models have emerged as a powerful tool in avian bioacoustics to assess environmental health. To maximize the potential of cost-effective and minimal-invasive passive acoustic monitoring (PAM), DL models must analyze bird vocalizations across a wide range of species and environmental conditions. However, data fragmentation challenges a comprehensive evaluation of generalization performance. Therefore, we introduce the BirdSet dataset, comprising approximately 520,000 global bird recordings for training and over 400 hours of PAM recordings for testing. Our benchmark offers baselines for several DL models to enhance comparability and consolidate research across studies, along with code implementations that include comprehensive training and evaluation protocols.
Mintlify uses AI to generate documentation from code – TechCrunch
Mintlify, a startup developing software to automate software documentation tasks, today announced that it raised $2.8 million in a seed round led by by Bain Capital Ventures with participation from TwentyTwo Ventures and Quinn Slack, Sourcegraph's co-founder. CEO Han Wang says that the proceeds will be put toward product development and doubling Mintlify's core, three-person team by the end of the year. Ithaca, New York-based Mintlify was co-founded in 2021 by Han Wang and Hahnbee Lee -- both software engineers by trade. Wang previously co-launched Foodful, an startup that developed a cloud-based monitoring system for cows, and Pe•ple, an online customer community platform that was acquired by Tribe in early 2021. Lee was a co-founder at Pe•ple before briefly joining Duolingo as an engineer.
This touch-sensitive glove is made from stretchy optical fibres
A touch-sensitive glove made from stretchable fibre-optic sensors could be used in robotics, sport and medicine. "We made a sensor that can sense haptic interactions, in the same way that our own skin sensors interact with [the] environment," says Hedan Bai at Cornell University in Ithaca, New York. Bai and her team created the glove using optical fibres made from thin elastomeric polyurethane cables that transmit light from an LED. The light is interrupted when the cables are bent, stretched or put under pressure. The team dyed parts of the fibres with different colours, meaning that as they are distorted, the colour of light coming out of the fibres changes.
Inner Workings: Crop researchers harness artificial intelligence to breed crops for the changing climate
Until recently, the field of plant breeding looked a lot like it did in centuries past. A breeder might examine, for example, which tomato plants were most resistant to drought and then cross the most promising plants to produce the most drought-resistant offspring. This process would be repeated, plant generation after generation, until, over the course of roughly seven years, the breeder arrived at what seemed the optimal variety. Researchers at ETH Zürich use standard color images and thermal images collected by drone to determine how plots of wheat with different genotypes vary in grain ripeness. Image credit: Norbert Kirchgessner (ETH Zürich, Zürich, Switzerland). Now, with the global population expected to swell to nearly 10 billion by 2050 (1) and climate change shifting growing conditions (2), crop breeder and geneticist Steven Tanksley doesn’t think plant breeders have that kind of time. “We have to double the productivity per acre of our major crops if we’re going to stay on par with the world’s needs,” says Tanksley, a professor emeritus at Cornell University in Ithaca, NY. To speed up the process, Tanksley and others are turning to artificial intelligence (AI). Using computer science techniques, breeders can rapidly assess which plants grow the fastest in a particular climate, which genes help plants thrive there, and which plants, when crossed, produce an optimum combination of genes for a given location, opting for traits that boost yield and stave off the effects of a changing climate. Large seed companies in particular have been using components of AI for more than a decade. With computing power rapidly advancing, the techniques are now poised to accelerate breeding on a broader scale. AI is not, however, a panacea. Crop breeders still grapple with tradeoffs such as higher yield versus marketable appearance. And even the most sophisticated AI …
Earphone cameras watch your facial expressions and read your lips
A wearable device consisting of two mini-cameras mounted on earphones can recognise your facial expressions and read your lips, even if your mouth is covered. The tool – called C-Face – was developed by Cheng Zhang at Cornell University in Ithaca, New York, and his colleagues. It looks at the sides of the wearer's head and uses machine learning to accurately visualise facial expressions by analysing small changes in cheek contour lines. "With previous technology to reconstruct facial expression, you had to put a camera in front of you. But that brings a lot of limitations," says Zhang. "Right now, many people are wearing a face mask, and standard facial tracking will not work. Our technology still works because it doesn't rely on what your face looks like."
Earphone cameras watch your facial expressions and read your lips
A wearable device consisting of two mini-cameras mounted on earphones can recognise your facial expressions and read your lips, even if your mouth is covered. The tool – called C-Face – was developed by Cheng Zhang at Cornell University in Ithaca, New York, and his colleagues. It looks at the sides of the wearer's head and uses machine learning to accurately visualise facial expressions by analysing small changes in cheek contour lines. "With previous technology to reconstruct facial expression, you had to put a camera in front of you. But that brings a lot of limitations," says Zhang. "Right now, many people are wearing a face mask, and standard facial tracking will not work. Our technology still works because it doesn't rely on what your face looks like."
Smiles beam and walls blush: Architecture meets AI at Microsoft
Jenny Sabin is perched high on a scissor lift, her head poking through an opening of the porous fabric structure that she's struggling to stretch onto the exoskeleton of her installation piece, which is suspended in the airy atrium of building 99 on Microsoft's Redmond, Washington, campus. Momentarily defeated, she pauses and looks up. "It's going to be gorgeous," she says. "It" is a glowing, translucent and ethereal pavilion that Sabin and her Microsoft collaborators describe as both a research tool and a glimpse into a future in which architecture and artificial intelligence merge. "To my knowledge, this installation is the first architectural structure to be driven by artificial intelligence in real time," said Sabin, principal designer at Jenny Sabin Studio in Ithaca, New York, who designed and built the pavilion as part of Microsoft's Artist in Residence program.